Goto

Collaborating Authors

 human-computer interaction


emg2pose: A Large and Diverse Benchmark for Surface Electromyographic Hand Pose Estimation

Neural Information Processing Systems

Hands are the primary means through which humans interact with the world. Reliable and always-available hand pose inference could yield new and intuitive control schemes for human-computer interactions, particularly in virtual and augmented reality. Computer vision is effective but requires one or multiple cameras and can struggle with occlusions, limited field of view, and poor lighting. Wearable wrist-based surface electromyography (sEMG) presents a promising alternative as an always-available modality sensing muscle activities that drive hand motion. However, sEMG signals are strongly dependent on user anatomy and sensor placement; existing sEMG models have thus required hundreds of users and device placements to effectively generalize for tasks other than pose inference. To facilitate progress on sEMG pose inference, we introduce the emg2pose benchmark, which is to our knowledge the first publicly available dataset of high-quality hand pose labels and wrist sEMG recordings.


How AI Is Reshaping Diplomacy and Global Affairs

TIME - Tech

With artificial intelligence putting productivity on hyperspeed, the painstaking but often slow nature of dealing with other countries, as well as policymaking, is also forced to speed up. But a panel at the forefront of these changes at the BRIDGE Summit in Abu Dhabi--which convenes creators, policymakers, investors, technologists, media institutions, and cultural leaders around the world to discuss the future of media--said that breaking things fast is not without consequences. "Decision makers are being asked to make decisions very quickly on the basis of information that may not be verified or verifiable," Elizabeth Churchill, a professor of Human-Computer Interaction from the Mohamed Bin Zayed University of Artificial Intelligence, told moderator Nikhil Kumar, an executive editor at TIME, which is a media partner of the BRIDGE Summit. Churchill, who held senior roles in firms like Google and Yahoo, said she returned to academia to explore transparent and "interrogable" AI tools and content that is effectively watermarked--so that decision-makers know at a glance if information is trustworthy. She said current shortfalls in information quality are "very much a design problem that sits at the surface of all of the tools that we use and in diplomacy conversations many different people are using."


Mutual Wanting in Human--AI Interaction: Empirical Evidence from Large-Scale Analysis of GPT Model Transitions

Shang, HaoYang, Liu, Xuan

arXiv.org Artificial Intelligence

The rapid evolution of large language models (LLMs) creates complex bidirectional expectations between users and AI systems that are poorly understood. We introduce the concept of "mutual wanting" to analyze these expectations during major model transitions. Through analysis of user comments from major AI forums and controlled experiments across multiple OpenAI models, we provide the first large-scale empirical validation of bidirectional desire dynamics in human-AI interaction. Our findings reveal that nearly half of users employ anthropomorphic language, trust significantly exceeds betrayal language, and users cluster into distinct "mutual wanting" types. We identify measurable expectation violation patterns and quantify the expectation-reality gap following major model releases. Using advanced NLP techniques including dual-algorithm topic modeling and multi-dimensional feature extraction, we develop the Mutual Wanting Alignment Framework (M-WAF) with practical applications for proactive user experience management and AI system design. These findings establish mutual wanting as a measurable phenomenon with clear implications for building more trustworthy and relationally-aware AI systems.


An Active Inference Model of Mouse Point-and-Click Behaviour

Klar, Markus, Stein, Sebastian, Paterson, Fraser, Williamson, John H., Murray-Smith, Roderick

arXiv.org Artificial Intelligence

We explore the use of Active Inference (AIF) as a computational user model for spatial pointing, a key problem in Human-Computer Interaction (HCI). We present an AIF agent with continuous state, action, and observation spaces, performing one-dimensional mouse pointing and clicking. We use a simple underlying dynamic system to model the mouse cursor dynamics with realistic perceptual delay. In contrast to previous optimal feedback control-based models, the agent's actions are selected by minimizing Expected Free Energy, solely based on preference distributions over percepts, such as observing clicking a button correctly. Our results show that the agent creates plausible pointing movements and clicks when the cursor is over the target, with similar end-point variance to human users. In contrast to other models of pointing, we incorporate fully probabilistic, predictive delay compensation into the agent. The agent shows distinct behaviour for differing target difficulties without the need to retune system parameters, as done in other approaches. We discuss the simulation results and emphasize the challenges in identifying the correct configuration of an AIF agent interacting with continuous systems.


Towards Human-Centered RegTech: Unpacking Professionals' Strategies and Needs for Using LLMs Safely

Hu, Siying, Yao, Yaxing, Lu, Zhicong

arXiv.org Artificial Intelligence

Large Language Models are profoundly changing work patterns in high-risk professional domains, yet their application also introduces severe and underexplored compliance risks. To investigate this issue, we conducted semi-structured interviews with 24 highly-skilled knowledge workers from industries such as law, healthcare, and finance. The study found that these experts are commonly concerned about sensitive information leakage, intellectual property infringement, and uncertainty regarding the quality of model outputs. In response, they spontaneously adopt various mitigation strategies, such as actively distorting input data and limiting the details in their prompts. However, the effectiveness of these spontaneous efforts is limited due to a lack of specific compliance guidance and training for Large Language Models. Our research reveals a significant gap between current NLP tools and the actual compliance needs of experts. This paper positions these valuable empirical findings as foundational work for building the next generation of Human-Centered, Compliance-Driven Natural Language Processing for Regulatory Technology (RegTech), providing a critical human-centered perspective and design requirements for engineering NLP systems that can proactively support expert compliance workflows.


Measuring and mitigating overreliance is necessary for building human-compatible AI

Ibrahim, Lujain, Collins, Katherine M., Kim, Sunnie S. Y., Reuel, Anka, Lamparth, Max, Feng, Kevin, Ahmad, Lama, Soni, Prajna, Kattan, Alia El, Stein, Merlin, Swaroop, Siddharth, Sucholutsky, Ilia, Strait, Andrew, Liao, Q. Vera, Bhatt, Umang

arXiv.org Artificial Intelligence

Large language models (LLMs) distinguish themselves from previous technologies by functioning as collaborative "thought partners," capable of engaging more fluidly in natural language. As LLMs increasingly influence consequential decisions across diverse domains from healthcare to personal advice, the risk of overreliance - relying on LLMs beyond their capabilities - grows. This position paper argues that measuring and mitigating overreliance must become central to LLM research and deployment. First, we consolidate risks from overreliance at both the individual and societal levels, including high-stakes errors, governance challenges, and cognitive deskilling. Then, we explore LLM characteristics, system design features, and user cognitive biases that - together - raise serious and unique concerns about overreliance in practice. We also examine historical approaches for measuring overreliance, identifying three important gaps and proposing three promising directions to improve measurement. Finally, we propose mitigation strategies that the AI research community can pursue to ensure LLMs augment rather than undermine human capabilities.


RelAItionship Building: Analyzing Recruitment Strategies for Participatory AI

Kim, Eugene, Balloli, Vaibhav, Karimian, Berelian, Bondi-Kelly, Elizabeth, Fish, Benjamin

arXiv.org Artificial Intelligence

Participatory AI, in which impacted community members and other stakeholders are involved in the design and development of AI systems, holds promise as a way to ensure AI is developed to meet their needs and reflect their values. However, the process of identifying, reaching out, and engaging with all relevant stakeholder groups, which we refer to as recruitment methodology, is still a practical challenge in AI projects striving to adopt participatory practices. In this paper, we investigate the challenges that researchers face when designing and executing recruitment methodology for Participatory AI projects, and the implications of current recruitment practice for Participatory AI. First, we describe the recruitment methodologies used in AI projects using a corpus of 37 projects to capture the diversity of practices in the field and perform an initial analysis on the documentation of recruitment practices, as well as specific strategies that researchers use to meet goals of equity and empowerment. To complement this analysis, we interview five AI researchers to learn about the outcomes of recruitment methodologies. We find that these outcomes are shaped by structural conditions of their work, researchers' own goals and expectations, and the relationships built from the recruitment methodology and subsequent collaboration. Based on these analyses, we provide recommendations for designing and executing relationship-forward recruitment methods, as well as reflexive recruitment documentation practices for Participatory AI researchers.


The Quasi-Creature and the Uncanny Valley of Agency: A Synthesis of Theory and Evidence on User Interaction with Inconsistent Generative AI

Manhaes, Mauricio, Miller, Christine, Schroeder, Nicholas

arXiv.org Artificial Intelligence

The user experience with large-scale generative AI is paradoxical: superhuman fluency meets absurd failures in common sense and consistency. This paper argues that the resulting potent frustration is an ontological problem, stemming from the "Quasi-Creature"-an entity simulating intelligence without embodiment or genuine understanding. Interaction with this entity precipitates the "Uncanny Valley of Agency," a framework where user comfort drops when highly agentic AI proves erratically unreliable. Its failures are perceived as cognitive breaches, causing profound cognitive dissonance. Synthesizing HCI, cognitive science, and philosophy of technology, this paper defines the Quasi-Creature and details the Uncanny Valley of Agency. An illustrative mixed-methods study ("Move 78," N=37) of a collaborative creative task reveals a powerful negative correlation between perceived AI efficiency and user frustration, central to the negative experience. This framework robustly explains user frustration with generative AI and has significant implications for the design, ethics, and societal integration of these powerful, alien technologies.


Toward AI Matching Policies in Homeless Services: A Qualitative Study with Policymakers

Johnston, Caroline M., Koumoundouros, Olga, Hwang, Angel Hsing-Chi, Onasch-Vera, Laura, Rice, Eric, Vayanos, Phebe

arXiv.org Artificial Intelligence

Artificial intelligence researchers have proposed various data-driven algorithms to improve the processes that match individuals experiencing homelessness to scarce housing resources. It remains unclear whether and how these algorithms are received or adopted by practitioners and what their corresponding consequences are. Through semi-structured interviews with 13 policymakers in homeless services in Los An-geles, we investigate whether such change-makers are open to the idea of integrating AI into the housing resource matching process, identifying where they see potential gains and drawbacks from such a system in issues of efficiency, fairness, and transparency. Our qualitative analysis indicates that, even when aware of various complicating factors, policymak-ers welcome the idea of an AI matching tool if thoughtfully designed and used in tandem with human decision-makers. Though there is no consensus as to the exact design of such an AI system, insights from policymakers raise open questions and design considerations that can be enlightening for future researchers and practitioners who aim to build responsible algorithmic systems to support decision-making in low-resource scenarios.


Emotion Detection Using Conditional Generative Adversarial Networks (cGAN): A Deep Learning Approach

Srivastava, Anushka

arXiv.org Artificial Intelligence

--Emotion recognition is a key task in affective computing with applications in healthcare, human-computer interaction, and surveillance systems. This study proposes a Conditional Generative Adversarial Network (cGAN)-based approach to generate synthetic emotion-specific facial images to augment training data and mitigate class imbalance. The generator learns to synthesize grayscale 64 64 facial images conditioned on emotion labels, while the discriminator distinguishes between real and generated images using label conditioning. The model was trained on the FER-2013 dataset and evaluated over 300 epochs. Training results demonstrate stable adversarial loss convergence, indicating effective learning and generation capability.